Stable Diffusion 3, with its novel Multimodal Diffusion Transformer architecture, surpasses leading text-to-image models by enhancing prompt comprehension and typography through separate processing weights for text and images, promising advancements in AI-generated visual content.
Wednesday, March 6, 2024Stable Diffusion 3 is a powerful image generation model. This paper introduces Latent Adversarial Diffusion Distillation, which reduces the number of diffusion steps down to 4 while maintaining image generation quality.
Stability AI has made its latest text-to-image AI model, Stable Diffusion 3, available to some developers via API and its new content creation platform called Stable Assistant Beta. The model is still in preview and not yet available to the general public.
Stability AI has released SD 3 weights for non-commercial use, with commercial use available on request.
Stable Diffusion 3 Medium is out. A cutting-edge text-to-image model that generates photorealistic images with 2 billion parameters, it overcomes common artifacts in hands and faces, handles complex prompts, and features enhanced typography. Despite recent legal and financial challenges, Stability AI continues to push the boundaries of generative AI, with future upgrades planned across video, audio, and language.